On the Equivalence between Canonical Correlation Analysis and Orthonormalized Partial Least Squares
نویسندگان
چکیده
Canonical correlation analysis (CCA) and partial least squares (PLS) are well-known techniques for feature extraction from two sets of multidimensional variables. The fundamental difference between CCA and PLS is that CCA maximizes the correlation while PLS maximizes the covariance. Although both CCA and PLS have been applied successfully in various applications, the intrinsic relationship between them remains unclear. In this paper, we attempt to address this issue by showing the equivalence relationship between CCA and orthonormalized partial least squares (OPLS), a variant of PLS. We further extend the equivalence relationship to the case when regularization is employed for both sets of variables. In addition, we show that the CCA projection for one set of variables is independent of the regularization on the other set of variables. We have performed experimental studies using both synthetic and real data sets and our results confirm the established equivalence relationship. The presented analysis provides novel insights into the connection between these two existing algorithms as well as the effect of the regularization.
منابع مشابه
Kernel PLS-SVC for Linear and Nonlinear Classification
A new method for classification is proposed. This is based on kernel orthonormalized partial least squares (PLS) dimensionality reduction of the original data space followed by a support vector classifier. Unlike principal component analysis (PCA), which has previously served as a dimension reduction step for discrimination problems, orthonormalized PLS is closely related to Fisher’s approach t...
متن کاملPLS classification of functional data
Partial least squares (PLS) approach is proposed for linear discriminant analysis (LDA) when predictors are data of functional type (curves). Based on the equivalence between LDA and the multiple linear regression (binary response) and LDA and the canonical correlation analysis (more than two groups), the PLS regression on functional data is used to estimate the discriminant coefficient functio...
متن کاملSparse Kernel Orthonormalized PLS for feature extraction in large data sets
In this paper we are presenting a novel multivariate analysis method for large scale problems. Our scheme is based on a novel kernel orthonormalized partial least squares (PLS) variant for feature extraction, imposing sparsity constrains in the solution to improve scalability. The algorithm is tested on a benchmark of UCI data sets, and on the analysis of integrated short-time music features fo...
متن کاملSparse Orthonormalized Partial Least Squares
Orthonormalized partial least squares (OPLS) is often used to find a low-rank mapping between inputs X and outputs Y by estimating loading matrices A and B. In this paper, we introduce sparse orthonormalized PLS as an extension of conventional PLS that finds sparse estimates of A through the use of the elastic net algorithm. We apply sparse OPLS to the reconstruction of presented images from BO...
متن کاملA Uniied Approach to Pca, Pls, Mlr and Cca
This paper presents a novel algorithm for analysis of stochastic processes. The algorithm can be used to nd the required solutions in the cases of principal component analysis (PCA), partial least squares (PLS), canonical correlation analysis (CCA) or multiple linear regression (MLR). The algorithm is iterative and sequential in its structure and uses on-line stochastic approximation to reach a...
متن کامل